Goto

Collaborating Authors

 vascular structure


A Novel Attention-Augmented Wavelet YOLO System for Real-time Brain Vessel Segmentation on Transcranial Color-coded Doppler

Zhang, Wenxuan, Li, Shuai, Wang, Xinyi, Sun, Yu, Kang, Hongyu, Wan, Pui Yuk Chryste, Qin, Jing, Zhang, Yuanpeng, Zheng, Yong-Ping, Lam, Sai-Kit

arXiv.org Artificial Intelligence

The Circle of Willis (CoW), vital for ensuring consistent blood flow to the brain, is closely linked to ischemic stroke. Accurate assessment of the CoW is important for identifying individuals at risk and guiding appropriate clinical management. Among existing imaging methods, Transcranial Color-coded Doppler (TCCD) offers unique advantages due to its radiation-free nature, affordability, and accessibility. However, reliable TCCD assessments depend heavily on operator expertise for identifying anatomical landmarks and performing accurate angle correction, which limits its widespread adoption. To address this challenge, we propose an AI-powered, real-time CoW auto-segmentation system capable of efficiently capturing cerebral arteries. No prior studies have explored AI-driven cerebrovascular segmentation using TCCD. In this work, we introduce a novel Attention-Augmented Wavelet YOLO (AAW-YOLO) network tailored for TCCD data, designed to provide real-time guidance for brain vessel segmentation in the CoW. We prospectively collected TCCD data comprising 738 annotated frames and 3,419 labeled artery instances to establish a high-quality dataset for model training and evaluation. The proposed AAW-YOLO demonstrated strong performance in segmenting both ipsilateral and contralateral CoW vessels, achieving an average Dice score of 0.901, IoU of 0.823, precision of 0.882, recall of 0.926, and mAP of 0.953, with a per-frame inference speed of 14.199 ms. This system offers a practical solution to reduce reliance on operator experience in TCCD-based cerebrovascular screening, with potential applications in routine clinical workflows and resource-constrained settings. Future research will explore bilateral modeling and larger-scale validation.


SWinMamba: Serpentine Window State Space Model for Vascular Segmentation

Zhao, Rongchang, Liu, Huanchi, Zhang, Jian

arXiv.org Artificial Intelligence

Vascular segmentation in medical images is crucial for disease diagnosis and surgical navigation. However, the segmented vascular structure is often discontinuous due to its slender nature and inadequate prior modeling. In this paper, we propose a novel Serpentine Window Mamba (SWinMamba) to achieve accurate vascular segmentation. The proposed SWinMamba innovatively models the continuity of slender vascular structures by incorporating serpentine window sequences into bidirectional state space models. The serpentine window sequences enable efficient feature capturing by adaptively guiding global visual context modeling to the vascular structure. Specifically, the Serpentine Window Tokenizer (SWToken) adaptively splits the input image using overlapping serpentine window sequences, enabling flexible receptive fields (RFs) for vascular structure modeling. The Bidirectional Aggregation Module (BAM) integrates coherent local features in the RFs for vascular continuity representation. In addition, dual-domain learning with Spatial-Frequency Fusion Unit (SFFU) is designed to enhance the feature representation of vascular structure. Extensive experiments on three challenging datasets demonstrate that the proposed SWinMamba achieves superior performance with complete and connected vessels.


AI-Powered Automated Model Construction for Patient-Specific CFD Simulations of Aortic Flows

Du, Pan, An, Delin, Wang, Chaoli, Wang, Jian-Xun

arXiv.org Artificial Intelligence

Effectively understanding and managing CVD requires advanced diagnostic tools capable of accurately characterizing complex hemodynamics within the cardiovascular system. While medical imaging modalities such as computed tomography (CT) and magnetic resonance imaging (MRI) provide high-resolution anatomical detail, they lack the capability to directly capture hemodynamics information (e.g., blood flow patterns, pressure, and wall shear stress fields) critical for understanding vascular function and pathology. To bridge this gap, image-based computational fluid dynamics (CFD) has emerged as a powerful computational paradigm that derives hemodynamic information from anatomical images via conservation laws. Although widely utilized in cardiovascular research, the clinical application of image-based CFD for diagnosis and surgical planning remains limited, largely due to the challenges associated with efficient and accurate model construction [2-4]. Constructing patient-specific vascular models for image-based CFD involves multiple steps, including image segmentation, geometry modeling, and mesh generation for the computational domain, all of which are critical to ensuring the fidelity of the final simulation results. However, the standard workflow heavily relies on manual methods, making it highly labor-intensive and time-consuming.


Large Language Model with Region-guided Referring and Grounding for CT Report Generation

Chen, Zhixuan, Bie, Yequan, Jin, Haibo, Chen, Hao

arXiv.org Artificial Intelligence

Computed tomography (CT) report generation is crucial to assist radiologists in interpreting CT volumes, which can be time-consuming and labor-intensive. Existing methods primarily only consider the global features of the entire volume, making it struggle to focus on specific regions and potentially missing abnormalities. To address this issue, we propose Reg2RG, the first region-guided referring and grounding framework for CT report generation, which enhances diagnostic performance by focusing on anatomical regions within the volume. Specifically, we utilize masks from a universal segmentation module to capture local features for each referring region. A local feature decoupling (LFD) strategy is proposed to preserve the local high-resolution details with little computational overhead. Then the local features are integrated with global features to capture inter-regional relationships within a cohesive context. Moreover, we propose a novel region-report alignment (RRA) training strategy. It leverages the recognition of referring regions to guide the generation of region-specific reports, enhancing the model's referring and grounding capabilities while also improving the report's interpretability. A large language model (LLM) is further employed as the language decoder to generate reports from integrated visual features, facilitating region-level comprehension. Extensive experiments on two large-scale chest CT-report datasets demonstrate the superiority of our method, which outperforms several state-of-the-art methods in terms of both natural language generation and clinical efficacy metrics while preserving promising interpretability. The code will be made publicly available.


KLDD: Kalman Filter based Linear Deformable Diffusion Model in Retinal Image Segmentation

Zhao, Zhihao, Zhao, Yinzheng, Yang, Junjie, Huang, Kai, Navab, Nassir, Nasseri, M. Ali

arXiv.org Artificial Intelligence

AI-based vascular segmentation is becoming increasingly common in enhancing the screening and treatment of ophthalmic diseases. Deep learning structures based on U-Net have achieved relatively good performance in vascular segmentation. However, small blood vessels and capillaries tend to be lost during segmentation when passed through the traditional U-Net downsampling module. To address this gap, this paper proposes a novel Kalman filter based Linear Deformable Diffusion (KLDD) model for retinal vessel segmentation. Our model employs a diffusion process that iteratively refines the segmentation, leveraging the flexible receptive fields of deformable convolutions in feature extraction modules to adapt to the detailed tubular vascular structures. More specifically, we first employ a feature extractor with linear deformable convolution to capture vascular structure information form the input images. To better optimize the coordinate positions of deformable convolution, we employ the Kalman filter to enhance the perception of vascular structures in linear deformable convolution. Subsequently, the features of the vascular structures extracted are utilized as a conditioning element within a diffusion model by the Cross-Attention Aggregation module (CAAM) and the Channel-wise Soft Attention module (CSAM). These aggregations are designed to enhance the diffusion model's capability to generate vascular structures. Experiments are evaluated on retinal fundus image datasets (DRIVE, CHASE_DB1) as well as the 3mm and 6mm of the OCTA-500 dataset, and the results show that the diffusion model proposed in this paper outperforms other methods.


Overcoming large 3D microscopy imaging data with AI

#artificialintelligence

Medical imaging generates an enormous amount of data that is impossible to manually analyze. With advances in artificial intelligence (AI) researchers are now looking at how this technology can be used to help manage and simplify the analysis of large 3D microscopy image datasets. In this SelectScience article, we speak with Jianxu Chen, head of the new Analysis of Microscopic BIOMedical Images (AMBIOM*) group at ISAS. Chen discusses how his group is developing scalable, AI-based image analysis algorithms to help support disease studies. Chen also explores some current trends in laboratories adopting AI and machine learning.